An Augmented Lagrangian Method for Non-Lipschitz Nonconvex Programming

نویسندگان

  • Xiaojun Chen
  • Lei Guo
  • Zhaosong Lu
  • Jane J. Ye
چکیده

We consider a class of constrained optimization problems where the objective function is a sum of a smooth function and a nonconvex non-Lipschitz function. Many problems in sparse portfolio selection, edge preserving image restoration and signal processing can be modelled in this form. First we propose the concept of the Karush-Kuhn-Tucker (KKT) stationary condition for the non-Lipschitz problem and show that it is necessary for optimality under a constraint qualification called the relaxed constant positive linear dependence (RCPLD) condition which is weaker than the Mangasarian-Fromovitz constraint qualification and holds automatically if all the constraint functions are affine. Then we propose an augmented Lagrangian method (AL) in which the augmented Lagrangian subproblems is solved by a non-monotone proximal gradient method. Under the assumption that a feasible point is known, we show that any accumulation point of the sequence generated by our method must be a feasible point. Moreover, if RCPLD holds at such an accumulation point, then it is a KKT point of the original problem. Finally we conduct numerical experiments to compare the performance of our AL method and the interior point (IP) method for solving two sparse portfolio selection models. The numerical results demonstrate that our method is not only comparable to the IP method in terms of solution quality, but also substantially faster than the IP method.

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

A Feasible Augmented Lagrangian Method for Non-lipschitz Nonconvex Programming

We consider a class of constrained optimization problems where the objective function is a sum of a smooth function and a nonconvex non-Lipschitz function. Many problems in sparse portfolio selection, edge preserving image restoration and signal processing can be modelled in this form. First we propose the concept of the Karush-Kuhn-Tucker (KKT) stationary condition for the non-Lipschitz proble...

متن کامل

Convergence Results on Proximal Method of Multipliers in Nonconvex Programming

We describe a primal-dual application of the proximal point algorithm to nonconvex minimization problems. Motivated by the work of Spingarn and more recently by the work of Kaplan and Tichatschke about the proximal point methodology in nonconvex optimization. This paper discusses some local results in two directions. The first one concerns the application of the proximal method of multipliers t...

متن کامل

An Augmented Lagrangian Based Algorithm for Distributed NonConvex Optimization

This paper is about distributed derivative-based algorithms for solving optimization problems with a separable (potentially nonconvex) objective function and coupled affine constraints. A parallelizable method is proposed that combines ideas from the fields of sequential quadratic programming and augmented Lagrangian algorithms. The method negotiates shared dual variables that may be interprete...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

عنوان ژورنال:
  • SIAM J. Numerical Analysis

دوره 55  شماره 

صفحات  -

تاریخ انتشار 2017